The road to future AI is paved with trust
- Date:
- September 30, 2022
- Source:
- Linköping University
- Summary:
- The place of artificial intelligence, AI, in our everyday life is increasing and many researchers believe that what we have seen so far is only the beginning. However, AI must be trustworthy in all situations. A new project has drawn up a research-based roadmap intended to guide research funding bodies and decision-makers towards the trustworthy AI of the future.
- Share:
The place of artificial intelligence, AI, in our everyday life is increasing and many researchers believe that what we have seen so far is only the beginning. However, AI must be trustworthy in all situations. Linköping University is coordinating TAILOR, a EU project that has drawn up a research-based roadmap intended to guide research funding bodies and decision-makers towards the trustworthy AI of the future.
"The development of artificial intelligence is in its infancy. When we look back at what we are doing today in 50 years, we will find it pretty primitive. In other words, most of the field remains to be discovered. That's why it's important to lay the foundation of trustworthy AI now," says Fredrik Heintz, professor of artificial intelligence at LiU, and coordinator of the TAILOR project.
TAILOR is one of six research networks set up by the EU to strengthen research capacity and develop the AI of the future. The foundation of trustworthy AI is being laid by TAILOR, by drawing up a framework, guidelines and a specification of the needs of the AI research community. "TAILOR" is an abbreviation of Foundations of Trustworthy AI -- integrating, learning, optimisation and reasoning.
The roadmap now presented by TAILOR is the first step on the way to standardisation, where the idea is that decision-makers and research funding bodies can gain insight into what is required to develop trustworthy AI. Fredrik Heintz believes that it is a good idea to show that many research problems must be solved before this can be achieved.
The researchers have defined three criteria for trustworthy AI: it must conform to laws and regulations, it must satisfy several ethical principles, and its implementation it must be robust and safe. Fredrik Heintz points out that these criteria pose major challenges, in particular the implementation of the ethical principles.
"Take justice, for example. Does this mean an equal distribution of resources or that all actors receive the resources needed to bring them all to the same level? We are facing major long-term questions, and it will take time before they are answered. Remember -- the definition of justice has been debated by philosophers and scholars for hundreds of years," says Fredrik Heintz.
The project will focus on large comprehensive research questions, and will attempt to find standards that all who work with AI can adopt. But Fredrik Heintz is convinced that we can only achieve this if basic research into AI is given priority.
"People often regard AI as a technology issue, but what's really important is whether we gain societal benefit from it. If we are to obtain AI that can be trusted and that functions well in society, we must make sure that it is centred on people," says Fredrik Heintz.
Many of the legal proposals written within the EU and its member states are written by legal specialists. But Fredrik Heintz believes that they lack expert knowledge within AI, which is a problem.
"Legislation and standards must be based on knowledge. This is where we researchers can contribute, providing information about the current forefront of research, and making well-grounded decisions possible. It's important that experts have the opportunity to influence questions of this type," says Fredrik Heintz.
The complete roadmap is available at: Strategic Research and Innovation Roadmap of trustworthy AI
Story Source:
Materials provided by Linköping University. Original written by Anders Törneholm. Note: Content may be edited for style and length.
Cite This Page: